在信息检索(IR)系统中,趋势和用户的兴趣可能会随着时间的推移而变化,改变要建议的请求或内容的分布。由于神经排名越来越依赖于培训数据,因此了解最近IR方法的转移能力在长期地址新域名的转移能力至关重要。在本文中,我们首先提出基于MSMarco语料库的数据集,旨在建模长期的主题以及IR属性驱动的受控设置。然后,我们深入分析最近神经红外模型的能力,同时不断地学习这些流。我们的实证研究突出显示在其中发生灾难性遗忘(例如,任务之间的相似程度,文本长度的特点,学习模型的方式),以便在模型设计方面提供未来的方向。
translated by 谷歌翻译
在这项工作中,我们的目标是提供自然语言的结构化答案,以便复杂的信息需求。特别是,我们从数据到文本生成的角度来设想使用生成模型。我们建议使用内容选择和规划管道,该管道旨在通过生成中间计划来构建答案。使用TREC复杂答案检索(CAR)数据集进行实验评估。我们评估生成的答案及其相应的结构,并显示了与文本到文本模型相比的基于规划的模型的有效性。
translated by 谷歌翻译
Anomaly detection on time series data is increasingly common across various industrial domains that monitor metrics in order to prevent potential accidents and economic losses. However, a scarcity of labeled data and ambiguous definitions of anomalies can complicate these efforts. Recent unsupervised machine learning methods have made remarkable progress in tackling this problem using either single-timestamp predictions or time series reconstructions. While traditionally considered separately, these methods are not mutually exclusive and can offer complementary perspectives on anomaly detection. This paper first highlights the successes and limitations of prediction-based and reconstruction-based methods with visualized time series signals and anomaly scores. We then propose AER (Auto-encoder with Regression), a joint model that combines a vanilla auto-encoder and an LSTM regressor to incorporate the successes and address the limitations of each method. Our model can produce bi-directional predictions while simultaneously reconstructing the original time series by optimizing a joint objective function. Furthermore, we propose several ways of combining the prediction and reconstruction errors through a series of ablation studies. Finally, we compare the performance of the AER architecture against two prediction-based methods and three reconstruction-based methods on 12 well-known univariate time series datasets from NASA, Yahoo, Numenta, and UCR. The results show that AER has the highest averaged F1 score across all datasets (a 23.5% improvement compared to ARIMA) while retaining a runtime similar to its vanilla auto-encoder and regressor components. Our model is available in Orion, an open-source benchmarking tool for time series anomaly detection.
translated by 谷歌翻译
The number of international benchmarking competitions is steadily increasing in various fields of machine learning (ML) research and practice. So far, however, little is known about the common practice as well as bottlenecks faced by the community in tackling the research questions posed. To shed light on the status quo of algorithm development in the specific field of biomedical imaging analysis, we designed an international survey that was issued to all participants of challenges conducted in conjunction with the IEEE ISBI 2021 and MICCAI 2021 conferences (80 competitions in total). The survey covered participants' expertise and working environments, their chosen strategies, as well as algorithm characteristics. A median of 72% challenge participants took part in the survey. According to our results, knowledge exchange was the primary incentive (70%) for participation, while the reception of prize money played only a minor role (16%). While a median of 80 working hours was spent on method development, a large portion of participants stated that they did not have enough time for method development (32%). 25% perceived the infrastructure to be a bottleneck. Overall, 94% of all solutions were deep learning-based. Of these, 84% were based on standard architectures. 43% of the respondents reported that the data samples (e.g., images) were too large to be processed at once. This was most commonly addressed by patch-based training (69%), downsampling (37%), and solving 3D analysis tasks as a series of 2D tasks. K-fold cross-validation on the training set was performed by only 37% of the participants and only 50% of the participants performed ensembling based on multiple identical models (61%) or heterogeneous models (39%). 48% of the respondents applied postprocessing steps.
translated by 谷歌翻译
强大的电力系统的长期计划需要了解不断变化的需求模式。电力需求对天气敏感。因此,引入间歇性可再生能源的供应方面变化与可变需求并列,将在网格计划过程中引入其他挑战。通过了解美国温度的空间和时间变化,可以分开需求对自然变异性和与气候变化相关的影响的需求的响应,尤其是因为尚不清楚由于前一个因素所产生的影响。通过该项目,我们旨在通过开发机器和深入学习“背面销售”模型来更好地支持电力系统的技术和政策开发过程,以重建多年需求记录并研究温度的自然变异性及其对需求的影响。
translated by 谷歌翻译
已显示来自各种来源的多模式信息的集成可以提高机器学习模型的性能,因此近年来受到了越来越多的关注。通常,这样的模型使用深度模式特异性网络来获得单峰特征,这些特征合并以获得“晚融合”表示。但是,这些设计承担了各自单峰管道中信息损失的风险。另一方面,结合早期特征的“早期融合”方法遭受了与特征异质性和高样本复杂性相关的问题。在这项工作中,我们提出了一种迭代表示的改进方法,称为渐进式融合,该方法减轻了晚期融合表示的问题。我们的模型不足的技术引入了向后连接,使后期融合的表示形式可用于早期层,从而提高了这些阶段的表示表现力,同时保留了晚期融合设计的优势。我们在任务上测试渐进式融合,包括情感检测,多媒体分析以及与不同模型的时间序列融合,以证明其多功能性。我们表明,我们的方法始终提高性能,例如,在多模式时间序列预测中,MSE降低了5%,鲁棒性提高了40%。
translated by 谷歌翻译
由于传感器的成本和可靠性高,泵的设计人员会尽可能地估算可行操作点所需的传感器数量。获得良好估计的主要挑战是可用的数据量低。使用此数量的数据,估算方法的性能不足以满足客户的要求。为了解决这个缺乏数据的问题,获取高质量数据对于获得良好的估计很重要。根据这些考虑,我们开发了一个主动学习框架,用于估计能量场中使用的模块化多泵的工作点。特别是,我们专注于电涌距离的估计。我们应用主动学习以使用最小数据集估算浪涌距离。结果报告说,主动学习也是真正应用的宝贵技术。
translated by 谷歌翻译
在许多现实世界应用中,可靠的概率估计在具有固有的不确定性的许多现实应用中至关重要,例如天气预报,医疗预后或自动车辆的碰撞避免。概率估计模型培训观察到的结果(例如,它是否已下雨,或者是否患者是否已死亡),因为感兴趣事件的地面真理概率通常是未知的。因此,问题类似于二进制分类,具有重要差异,即目标是估计概率而不是预测特定结果。这项工作的目标是使用深神经网络调查从高维数据的概率估计。存在几种方法来改善这些模型产生的概率,但它们主要专注于概率与模型不确定性相关的分类问题。在具有固有的不确定性问题的情况下,在没有访问地面概率的情况下评估性能有挑战性。要解决此问题,我们构建一个合成数据集以学习和比较不同的可计算度量。我们评估了合成数据以及三个现实世界概率估计任务的现有方法,所有这些方法都涉及固有的不确定性:从雷达图像的降水预测,从组织病理学图像预测癌症患者存活,并预测从Dashcam视频预测车祸。最后,我们还提出了一种使用神经网络的概率估计的新方法,该方法修改了培训过程,促进了与从数据计算的经验概率一致的输出概率。该方法优于模拟和真实数据上大多数度量的现有方法。
translated by 谷歌翻译